首页> 外文OA文献 >Revisiting Unreasonable Effectiveness of Data in Deep Learning Era
【2h】

Revisiting Unreasonable Effectiveness of Data in Deep Learning Era

机译:重温深度学习时代数据的不合理有效性

代理获取
本网站仅为用户提供外文OA文献查询和代理获取服务,本网站没有原文。下单后我们将采用程序或人工为您竭诚获取高质量的原文,但由于OA文献来源多样且变更频繁,仍可能出现获取不到、文献不完整或与标题不符等情况,如果获取不到我们将提供退款服务。请知悉。

摘要

The success of deep learning in vision can be attributed to: (a) models withhigh capacity; (b) increased computational power; and (c) availability oflarge-scale labeled data. Since 2012, there have been significant advances inrepresentation capabilities of the models and computational capabilities ofGPUs. But the size of the biggest dataset has surprisingly remained constant.What will happen if we increase the dataset size by 10x or 100x? This papertakes a step towards clearing the clouds of mystery surrounding therelationship between `enormous data' and visual deep learning. By exploitingthe JFT-300M dataset which has more than 375M noisy labels for 300M images, weinvestigate how the performance of current vision tasks would change if thisdata was used for representation learning. Our paper delivers some surprising(and some expected) findings. First, we find that the performance on visiontasks increases logarithmically based on volume of training data size. Second,we show that representation learning (or pre-training) still holds a lot ofpromise. One can improve performance on many vision tasks by just training abetter base model. Finally, as expected, we present new state-of-the-artresults for different vision tasks including image classification, objectdetection, semantic segmentation and human pose estimation. Our sincere hope isthat this inspires vision community to not undervalue the data and developcollective efforts in building larger datasets.
机译:视觉深度学习的成功可以归因于:(a)具有高能力的模型; (b)提高计算能力; (c)大规模标签数据的可用性。自2012年以来,模型的表示能力和GPU的计算能力有了长足的进步。但是最大的数据集的大小却令人惊讶地保持不变,如果我们将数据集的大小增加10倍或100倍会发生什么?本文迈出了一步,清除了“巨大数据”与视觉深度学习之间的关系之谜。通过利用JFT-300M数据集(其中包含超过375M的噪声标签)来提取300M图像,我们研究了如果将此数据用于表示学习,当前视觉任务的性能将如何变化。我们的论文提供了一些令人惊讶的(和一些预期的)发现。首先,我们发现视觉任务的性能基于训练数据量的大小呈对数增长。其次,我们证明了表征学习(或预训练)仍然有很多承诺。仅通过训练更好的基本模型,就可以提高许多视觉任务的性能。最后,正如预期的那样,我们为不同的视觉任务提供了最新的技术成果,包括图像分类,目标检测,语义分割和人体姿势估计。我们真诚的希望是,这可以激发视觉界不要低估数据并在建立更大的数据集方面进行集体努力。

著录项

相似文献

  • 外文文献
  • 中文文献
  • 专利
代理获取

客服邮箱:kefu@zhangqiaokeyan.com

京公网安备:11010802029741号 ICP备案号:京ICP备15016152号-6 六维联合信息科技 (北京) 有限公司©版权所有
  • 客服微信

  • 服务号